Microsoft Unveils On-Device AI Model ’Mu’ for Efficient Language Processing
Microsoft has introduced Mu, a compact language model designed to operate directly on devices like Copilot+ PCs without cloud dependency. Leveraging Neural Processing Units, Mu delivers rapid responses while optimizing power and memory usage—a stark contrast to bulkier cloud-based AI systems.
The model builds on Microsoft's Phi research, trained on high-quality educational data and refined through advanced techniques like distillation and low-rank adaptation. Its encoder-decoder architecture separates input/output processing, enabling greater efficiency through features like rotary positional embeddings and grouped-query attention.
Quantization techniques further enhance performance by simplifying calculations without compromising accuracy. This development signals a shift toward decentralized AI capabilities in consumer hardware.